Goto

Collaborating Authors

 icub robot


Learning secondary tool affordances of human partners using iCub robot's egocentric data

Ding, Bosong, Oztop, Erhan, Spigler, Giacomo, Kirtay, Murat

arXiv.org Artificial Intelligence

Objects, in particular tools, provide several action possibilities to the agents that can act on them, which are generally associated with the term of affordances. A tool is typically designed for a specific purpose, such as driving a nail in the case of a hammer, which we call as the primary affordance. A tool can also be used beyond its primary purpose, in which case we can associate this auxiliary use with the term secondary affordance. Previous work on affordance perception and learning has been mostly focused on primary affordances. Here, we address the less explored problem of learning the secondary tool affordances of human partners. To do this, we use the iCub robot to observe human partners with three cameras while they perform actions on twenty objects using four different tools. In our experiments, human partners utilize tools to perform actions that do not correspond to their primary affordances. For example, the iCub robot observes a human partner using a ruler for pushing, pulling, and moving objects instead of measuring their lengths. In this setting, we constructed a dataset by taking images of objects before and after each action is executed. We then model learning secondary affordances by training three neural networks (ResNet-18, ResNet-50, and ResNet-101) each on three tasks, using raw images showing the `initial' and `final' position of objects as input: (1) predicting the tool used to move an object, (2) predicting the tool used with an additional categorical input that encoded the action performed, and (3) joint prediction of both tool used and action performed. Our results indicate that deep learning architectures enable the iCub robot to predict secondary tool affordances, thereby paving the road for human-robot collaborative object manipulation involving complex affordances.


Utilization of Non-verbal Behaviour and Social Gaze in Classroom Human-Robot Interaction Communications

Shaghaghi, Sahand, Aliasghari, Pourya, Tripp, Bryan, Dautenhahn, Kerstin, Nehaniv, Chrystopher

arXiv.org Artificial Intelligence

This abstract explores classroom Human-Robot Interaction (HRI) scenarios with an emphasis on the adaptation of human-inspired social gaze models in robot cognitive architecture to facilitate a more seamless social interaction. First, we detail the HRI scenarios explored by us in our studies followed by a description of the social gaze model utilized for our research. We highlight the advantages of utilizing such an attentional model in classroom HRI scenarios. We also detail the intended goals of our upcoming study involving this social gaze model.


Human Impression of Humanoid Robots Mirroring Social Cues

Fu, Di, Abawi, Fares, Allgeuer, Philipp, Wermter, Stefan

arXiv.org Artificial Intelligence

Mirroring non-verbal social cues such as affect or movement can enhance human-human and human-robot interactions in the real world. The robotic platforms and control methods also impact people's perception of human-robot interaction. However, limited studies have compared robot imitation across different platforms and control methods. Our research addresses this gap by conducting two experiments comparing people's perception of affective mirroring between the iCub and Pepper robots and movement mirroring between vision-based iCub control and Inertial Measurement Unit (IMU)-based iCub control. We discovered that the iCub robot was perceived as more humanlike than the Pepper robot when mirroring affect. A vision-based controlled iCub outperformed the IMU-based controlled one in the movement mirroring task. Our findings suggest that different robotic platforms impact people's perception of robots' mirroring during HRI. The control method also contributes to the robot's mirroring performance. Our work sheds light on the design and application of different humanoid robots in the real world.


Real-time Addressee Estimation: Deployment of a Deep-Learning Model on the iCub Robot

Mazzola, Carlo, Rea, Francesco, Sciutti, Alessandra

arXiv.org Artificial Intelligence

Aiming at implementing AE skills in robots to let them Focusing on the perceptual domain, i.e., a passive agent interact in unstructured scenarios, this paper 1) describes the listening to humans, the artificial agents must be able to development of an AE deep-learning model trained on humanrobot detect voices (Sound Detection and Voice Recognition), recognize interaction (HRI) dataset, as already described in [16], 2) who is talking (Speaker Recognition and Speaker illustrates its first deployment on the humanoid robot iCub, and Localization), and what they are saying (Natural Language 3) reports the results of an HRI pilot experiment to evaluate Understanding). But even considering optimal performances in the performance of the model deployed on the iCub compared all these tasks, an artificial agent endowed with such abilities to previous tests made on the training dataset.


The Robot in the Room: Influence of Robot Facial Expressions and Gaze on Human-Human-Robot Collaboration

Fu, Di, Abawi, Fares, Wermter, Stefan

arXiv.org Artificial Intelligence

Robot facial expressions and gaze are important factors for enhancing human-robot interaction (HRI), but their effects on human collaboration and perception are not well understood, for instance, in collaborative game scenarios. In this study, we designed a collaborative triadic HRI game scenario, where two participants worked together to insert objects into a shape sorter. One participant assumed the role of a guide. The guide instructed the other participant, who played the role of an actor, by placing occluded objects into the sorter. A humanoid robot issued instructions, observed the interaction, and displayed social cues to elicit changes in the two participants' behavior. We measured human collaboration as a function of task completion time and the participants' perceptions of the robot by rating its behavior as intelligent or random. Participants also evaluated the robot by filling out the Godspeed questionnaire. We found that human collaboration was higher when the robot displayed a happy facial expression at the beginning of the game compared to a neutral facial expression. We also found that participants perceived the robot as more intelligent when it displayed a positive facial expression at the end of the game. The robot's behavior was also perceived as intelligent when directing its gaze toward the guide at the beginning of the interaction, not the actor. These findings provide insights into how robot facial expressions and gaze influence human behavior and perception in collaboration.


iCub! Do you recognize what I am doing?: multimodal human action recognition on multisensory-enabled iCub robot

Kniesmeijer, Kas, Kirtay, Murat

arXiv.org Artificial Intelligence

This study uses multisensory data (i.e., color and depth) to recognize human actions in the context of multimodal human-robot interaction. Here we employed the iCub robot to observe the predefined actions of the human partners by using four different tools on 20 objects. We show that the proposed multimodal ensemble learning leverages complementary characteristics of three color cameras and one depth sensor that improves, in most cases, recognition accuracy compared to the models trained with a single modality. The results indicate that the proposed models can be deployed on the iCub robot that requires multimodal action recognition, including social tasks such as partner-specific adaptation, and contextual behavior understanding, to mention a few.


Judging by the Look: The Impact of Robot Gaze Strategies on Human Cooperation

Fu, Di, Abawi, Fares, Strahl, Erik, Wermter, Stefan

arXiv.org Artificial Intelligence

Human eye gaze plays an important role in delivering information, communicating intent, and understanding others' mental states. Previous research shows that a robot's gaze can also affect humans' decision-making and strategy during an interaction. However, limited studies have trained humanoid robots on gaze-based data in human-robot interaction scenarios. Considering gaze impacts the naturalness of social exchanges and alters the decision process of an observer, it should be regarded as a crucial component in human-robot interaction. To investigate the impact of robot gaze on humans, we propose an embodied neural model for performing human-like gaze shifts. This is achieved by extending a social attention model and training it on eye-tracking data, collected by watching humans playing a game. We will compare human behavioral performances in the presence of a robot adopting different gaze strategies in a human-human cooperation game.


Self-supervised reinforcement learning for speaker localisation with the iCub humanoid robot

Gonzalez-Billandon, Jonas, Grasse, Lukas, Tata, Matthew, Sciutti, Alessandra, Rea, Francesco

arXiv.org Artificial Intelligence

In the future robots will interact more and more with humans and will have to communicate naturally and efficiently. Automatic speech recognition systems (ASR) will play an important role in creating natural interactions and making robots better companions. Humans excel in speech recognition in noisy environments and are able to filter out noise. Looking at a person's face is one of the mechanisms that humans rely on when it comes to filtering speech in such noisy environments. Having a robot that can look toward a speaker could benefit ASR performance in challenging environments. To this aims, we propose a self-supervised reinforcement learning-based framework inspired by the early development of humans to allow the robot to autonomously create a dataset that is later used to learn to localize speakers with a deep learning network.


Are we done with object recognition? The iCub robot's perspective

Pasquale, Giulia, Ciliberto, Carlo, Odone, Francesca, Rosasco, Lorenzo, Natale, Lorenzo

arXiv.org Artificial Intelligence

We report on an extensive study of the benefits and limitations of current deep learning approaches to object recognition in robot vision scenarios, introducing a novel dataset used for our investigation. To avoid the biases in currently available datasets, we consider a natural human-robot interaction setting to design a data-acquisition protocol for visual object recognition on the iCub humanoid robot. Analyzing the performance of off-the-shelf models trained off-line on large-scale image retrieval datasets, we show the necessity for knowledge transfer. We evaluate different ways in which this last step can be done, and identify the major bottlenecks affecting robotic scenarios. By studying both object categorization and identification problems, we highlight key differences between object recognition in robotics applications and in image retrieval tasks, for which the considered deep learning approaches have been originally designed. In a nutshell, our results confirm the remarkable improvements yield by deep learning in this setting, while pointing to specific open challenges that need be addressed for seamless deployment in robotics.


Robots of the future will learn like children and teach themselves

#artificialintelligence

European researchers aim to make independent robots which learn like children, setting their own goals and being curious about the world around them. The views expressed in the contents above are those of our users and do not necessarily reflect the views of MailOnline. By posting your comment you agree to our house rules.